STUDIO14: Agents, AI, and Azure AI Foundry - A Deep Dive into Enterprise AI Platforms

Session Date: May 19-22, 2025
Duration: ~45 minutes
Venue: Microsoft Build 2025 (Online)
Speakers:

  • Seth Juarez (Principal Program Manager, Microsoft) - Moderator
  • Marco Casalaina (VP Products & AI Futurist, Core AI, Microsoft)
  • Pablo Castro (CVP & Distinguished Engineer, AI Platform, Microsoft)
  • Yina Arenas (VP Azure AI Foundry, Microsoft)

Link: Microsoft Build 2025 Session STUDIO14


Table of Contents

  1. Introduction and Speaker Presentations
  2. The Data Foundation: AI as Language Calculators
  3. Advanced Document Processing and AI-Enhanced Indexing
  4. Evolution Beyond RAG: Agentic Retrieval
  5. The Model Explosion: Azure AI Foundry’s 10,000+ Models
  6. Defining Agents: LLMs as Control Flow
  7. Emerging Protocols and Standards
  8. Safety, Security, and Quality Assurance
  9. Agent Service: Production-Ready AI Microservices
  10. Real-World Customer Implementation: BMW Case Study
  11. Platform Integration and API Management

1. Introduction and Speaker Presentations

Timeframe: 00:00:00
Duration: 3m 15s
Speakers: Seth Juarez, Marco Casalaina, Pablo Castro, Yina Arenas

The session opened with Seth Juarez as moderator introducing the panel of AI experts from Microsoft. Each speaker provided their background and role within Microsoft’s AI ecosystem, setting the stage for a comprehensive discussion about enterprise AI platforms and agentic systems.

Key Introductions:

  • Marco Casalaina: VP Products of Core AI, self-described favorite color as “AI”
  • Yina Arenas: Leads Product for Azure AI Foundry, focusing on Build announcements
  • Pablo Castro: CVP in AI team, running the Azure AI Search team
  • Seth Juarez: Principal Program Manager, moderator and AI enthusiast

This introductory segment established the collaborative nature of the Microsoft AI teams and their shared focus on making AI accessible and powerful for enterprise customers.


2. The Data Foundation: AI as Language Calculators

Timeframe: 00:03:15
Duration: 7m 45s
Speakers: Seth Juarez, Pablo Castro

2.1. The Core Data Discipline Challenge

Seth Juarez introduced a fundamental principle that shaped the entire discussion: “AI only amplifies the data estate that you have. If you are not disciplined with your data, AI is going to make you seem not disciplined.”

This concept established the foundation for understanding why data quality is critical in AI implementations. The discussion revealed that:

  • AI amplifies existing practices - both good and bad data management becomes magnified
  • Quality determines outcomes - poor data discipline leads to poor AI results
  • Models as calculators - Seth’s analogy of “models are language calculators” emphasizes that incorrect inputs produce incorrect outputs

2.2. Azure AI Search’s Role in Data Quality

Pablo Castro explained how Azure AI Search addresses the data challenge by focusing on retrieval accuracy: “Our job from the retrieval systems perspective is, at every point in time, find you the right bit of information so the model has the information to know what to do next.”

The critical insight is that models will confidently provide incorrect answers when given poor data, making the retrieval system’s accuracy essential for AI success.


3. Advanced Document Processing and AI-Enhanced Indexing

Timeframe: 00:07:45
Duration: 12m 30s
Speakers: Pablo Castro, Seth Juarez

3.1. Document Cracking Innovation

Pablo Castro detailed Azure AI Search’s approach to handling diverse data formats through “document cracking” - an AI-powered process that:

  • Handles multiple formats - PDFs, images, text, and mixed media documents
  • Understands layout - AI extracts structural information and maintains context
  • Processes visual content - pictures and diagrams become searchable content
  • Optimizes chunking - intelligent document segmentation for better retrieval

3.2. AI-Powered Indexing Pipeline

The discussion revealed a comprehensive processing stack where users can “point us at your data, and if you don’t want to have an opinion, let’s just do it - we got you.” This approach includes:

  • Automatic vectorization - semantic embeddings without manual intervention
  • Layout preservation - document structure maintained for contextual understanding
  • Multi-modal processing - text, images, and structure processed together
  • Flexibility options - full automation or custom configuration based on user preferences

4. Evolution Beyond RAG: Agentic Retrieval

Timeframe: 00:12:30
Duration: 17m 20s
Speakers: Pablo Castro, Yina Arenas, Seth Juarez

4.1. Traditional RAG Limitations

Yina Arenas prompted the discussion about moving beyond traditional RAG (Retrieval-Augmented Generation) by noting: “We’ve been doing RAG for what? Two years now? But now we have new capabilities.”

The conversation acknowledged that while RAG has been successful, the accumulated learning over two years has revealed opportunities for more sophisticated approaches.

4.2. Agentic Retrieval Breakthrough

Pablo Castro announced a significant innovation: “We apply the same agentic methods we use in many other parts of the systems… to the search stack. So we have this agentic retrieval capability that we’re just rolling out.”

This breakthrough enables:

  • Reflective analysis - AI evaluates retrieval quality and completeness
  • Dynamic query expansion - automatic query reformulation and branching
  • Context-aware iteration - continuous improvement of search results
  • Multi-step reasoning - complex information gathering strategies

4.3. Advanced Ranking and Re-ranking

The discussion covered deep learning integration in search systems:

  • Transformer-based re-ranking - modern neural models for relevance scoring
  • Scale management - filtering from millions of documents to the top 3-5 results
  • Semantic understanding - meaning-based rather than keyword-based ranking
  • Developer simplicity - automated optimization with full semantic ranking stack

5. The Model Explosion: Azure AI Foundry’s 10,000+ Models

Timeframe: 00:17:20
Duration: 22m 45s
Speakers: Yina Arenas, Seth Juarez

5.1. Scale of Model Ecosystem

Yina Arenas revealed the dramatic growth in available AI models: “Two years ago, we had the OpenAI first three models… Now we have an explosion of models that has used the ecosystem. We are, as of today, have more than 10,000 models in the Azure AI Foundry catalog.”

This massive expansion includes:

  • Comprehensive coverage - text, speech, image, video processing models
  • Industry specialization - healthcare, finance, retail-specific models
  • Domain expertise - models optimized for specific use cases and scenarios

5.2. Model Selection and Discovery

To address the complexity of choosing from 10,000+ models, Azure AI Foundry provides:

  • Catalog organization - multiple ways to categorize and filter models
  • Leaderboard comparisons - cost, throughput, safety, and quality metrics
  • Scenario-based filtering - reasoning, text processing, image analysis categories
  • Use-case optimization - tools to find the right model for specific applications

5.3. Model Router: Intelligent Automation

Yina introduced the Model Router as a solution for automatic model selection: “Model router is an overlay on top of the set of models that you have deployed from Azure OpenAI, and what it will do is, based on the prompt, it will decide which model to use.”

Benefits include:

  • Cost optimization - simple prompts routed to cheaper Nano models
  • Performance matching - complex reasoning tasks routed to advanced models like O3
  • Developer simplicity - no manual model selection required
  • Transparent operation - seamless routing without code changes

6. Defining Agents: LLMs as Control Flow

Timeframe: 00:22:45
Duration: 28m 30s
Speakers: Yina Arenas, Seth Juarez

6.1. Simple Agent Definition

Yina Arenas provided a clear, developer-focused definition: “Agents is where you let a language model help you decide the control flow of the program.”

Seth Juarez expanded this with a programming analogy, suggesting agents represent a new control structure - like “if”, “while”, or “switch” statements - which he humorously called the “swift statement” where LLMs drive program flow control.

6.2. The Adaptability Breakthrough

The discussion contrasted traditional automation with AI-powered systems:

Traditional Automation Challenges:

  • Zero adaptability - hard-coded workflows break with any changes
  • Maintenance nightmare - constant script updates for workflow modifications
  • Rigid interactions - users saying “representative, representative, representative”

AI-Powered Flexibility:

  • Dynamic adaptation - AI adjusts to changing requirements automatically
  • Natural interaction - conversational interfaces replace rigid menu systems
  • Planning capability - AI can strategize multi-step approaches
  • Learning integration - systems improve through continued interaction

6.3. Function Calling and Tool Integration

Yina explained how agents gain capabilities through function calling: “They have the ability to do function calling, to call a tool, whether that is retrieving knowledge from Azure AI Search or making an action in a system, anything that can be described with an API can be called by the LLM.”

This enables:

  • Universal tool access - any API-describable functionality becomes available
  • Knowledge retrieval - integration with Azure AI Search for information access
  • Action execution - real system interactions and modifications
  • Composable functionality - mixing and matching tools for complex workflows

7. Emerging Protocols and Standards

Timeframe: 00:28:30
Duration: 32m 15s
Speakers: Yina Arenas, Seth Juarez

7.1. The Standards Evolution Challenge

Yina contextualized the current state of agent protocols: “It’s early in the development of the technologies around agentic AI, and we’ll take you back to the days before we had HTTP. We don’t have standards right now. They’re evolving.”

This creates challenges similar to early internet protocols, where multiple competing standards are emerging and market forces will determine which gain widespread adoption.

7.2. Key Protocol Support

The discussion covered essential protocols for agent systems:

Agent-to-Agent Communication:

  • A2A Protocol - standardized agent interaction and coordination
  • Multi-agent orchestration - complex workflow coordination capabilities

Tool Integration Standards:

  • Model Context Protocol (MCP) - standardized tool calling interface
  • OpenAPI integration - standard REST API connectivity
  • Custom protocol support - flexibility for proprietary systems

7.3. Azure AI Foundry Protocol Coverage

Yina emphasized Microsoft’s comprehensive approach: “In our offering in Azure AI Foundry, in the agent service, we support A2A, MCP, we support Assistants API, Responsys API. We’re working with LangChain and CrewAI to support their agent, their agentic API protocols as well.”

This “bring your stuff” philosophy ensures compatibility with existing toolchains and emerging standards.


8. Safety, Security, and Quality Assurance

Timeframe: 00:32:15
Duration: 36m 00s
Speakers: Yina Arenas, Seth Juarez

8.1. Multi-Dimensional Safety Approach

Yina outlined Azure AI Foundry’s comprehensive safety framework: “It’s not just about the models. It’s about the entire set of development environment that Azure AI Foundry offers for you.”

The quality evaluation framework includes:

  • Traditional evaluators - relevance and accuracy assessment
  • Agent-specific evaluation - tool calling correctness validation
  • Intent understanding - system prompt adherence monitoring
  • Instruction following - behavioral compliance assessment

8.2. Security and Protection

The security dimension encompasses:

  • Prompt shields - protection against injection attacks
  • Attack response systems - automated defense against adversarial inputs
  • Continuous monitoring - real-time security assessment
  • Quality optimization - ongoing performance improvement

8.3. Development Environment Integration

Azure AI Foundry integrates safety features throughout the development process:

  • Built-in evaluators - integrated quality assessment tools
  • Security umbrella - comprehensive protection across applications
  • Monitoring dashboards - real-time safety and performance metrics
  • Iterative improvement - continuous optimization workflows

9. Agent Service: Production-Ready AI Microservices

Timeframe: 00:36:00
Duration: 39m 30s
Speakers: Yina Arenas, Seth Juarez

9.1. General Availability Announcement

Yina announced a significant milestone: “Agent service is an offering that we are taking to general availability today at Build, and it is basically a very simple way for you to create your agent and run it on the cloud.”

This represents Azure’s commitment to making enterprise-grade agent deployment accessible without infrastructure complexity.

9.2. Agent Configuration Components

The service enables declarative agent definition through simple configuration:

Core Components:

  • Agent Identity - name and personality definition
  • Instructions - behavioral and operational guidelines
  • Tool Integration - APIs, functions, and services
  • Data Sources - Azure AI Search, Fabric, SharePoint connectivity
  • Knowledge Access - Bing integration for world knowledge
  • Action Capabilities - Logic Apps, Azure Functions, OpenAPI support

9.3. Integration Ecosystem

The comprehensive connectivity includes:

  • Azure services - native integration with Microsoft ecosystem
  • Third-party APIs - OpenAPI standard support
  • MCP servers - standardized tool protocol support
  • Custom functions - Azure Functions for specialized logic
  • Enterprise data - Fabric, SharePoint, and other data platforms

Seth characterized this as “an AI agentic microservice thing” - emphasizing the cloud-native, scalable nature of the service.


10. Real-World Customer Implementation: BMW Case Study

Timeframe: 00:39:30
Duration: 43m 45s
Speakers: Marco Casalaina, Seth Juarez

10.1. The BMW Sensor Data Challenge

Marco Casalaina shared a compelling customer success story from his visit to BMW in Munich: “They have these cars all over the world… they have all this special paint on these cars, but the cars have these sensors in them. It’s called ‘MDR,’ mobile data recorder, and these sensors record everything.”

The scale of the challenge:

  • 5,000 sensors per vehicle - comprehensive data collection
  • Global fleet monitoring - worldwide data aggregation
  • Multi-dimensional data - engine, brake, ambient temperature, moisture
  • Azure cloud storage - centralized data lake architecture

10.2. The Access Problem

The implementation faced a classic enterprise data challenge: “Nobody was able to query that. They put a lot of this stuff in a Kusto database, a SQL database, and nobody knew how to query that at BMW. So there was this special class of wizards who were the only people who could query these things.”

This created significant organizational barriers:

  • Expert dependency - only specialized “wizards” could access data
  • Cryptic naming - sensors with names like “Q underscore RSTR” were meaningless
  • Technical complexity - Kusto and SQL knowledge requirements
  • Knowledge bottleneck - limited organizational data utilization

10.3. AI-Powered Solution

BMW’s solution required significant upfront investment: “The folks at BMW, they spent six months creating a semantic model for this, because if you have a sensor called ‘Q underscore RSTR’ – nobody knows, not your AI and not you either.”

The implementation process included:

  • Semantic modeling - comprehensive sensor definition project
  • Cross-team collaboration - organization-wide knowledge gathering
  • Documentation creation - sensor names, purposes, ranges, and relationships
  • Azure AI integration - data agent implementation using Azure services

10.4. Transformative Results

The results demonstrated the power of democratized data access: “Now pretty much anybody at BMW could just query this stuff, and so could I. They actually let me use it… I said, ‘Show me all the hard-braking events in the last week in rainy weather,’ and it can totally pull that off.”

This transformation delivered:

  • Universal access - anyone at BMW can query sensor data
  • Natural language interface - no SQL knowledge required
  • Real-time insights - immediate answers to complex questions
  • Scalable knowledge - AI agent handles unlimited query complexity

11. Platform Integration and API Management

Timeframe: 00:43:45
Duration: 45m 00s
Speakers: Marco Casalaina, Seth Juarez

11.1. Live Demo: Travel Agent API Integration

Marco demonstrated the power of Azure’s integrated platform approach: “Yesterday, the need came up to connect this to an API, a flight reservation API. We’ve been working with this travel agent.”

The development workflow showcased:

  • Rapid requirement response - addressing same-day integration needs
  • Service discovery - leveraging Azure API Management for the first time
  • Mock API creation - rapid prototyping and development capabilities
  • Standard protocol exposure - automatic OpenAPI generation

11.2. Multi-Protocol API Exposure

The demo highlighted flexible integration options: “I cannot just create the API. I could expose it to the OpenAPI protocol. So instantly, just like this, I had an API that my agent can use, and now if I want to, I could just create a new MCP server for it also.”

This enables:

  • OpenAPI standard - immediate agent compatibility
  • MCP server creation - standardized protocol support
  • Multiple exposure methods - different protocols for different needs
  • Platform integration - Azure services working together seamlessly

11.3. The Integration Philosophy

Marco’s concluding thoughts emphasized practical necessity: “It may not be the sexiest thing in the world, but you need this stuff to be able to connect your agents.”

Seth summarized the comprehensive platform approach: “So we did the data, we did the models, and then you’re like – and the platform Azure makes it all good.”

This philosophy demonstrates:

  • Platform cohesion - Azure services designed to work together
  • Developer productivity - rapid integration without complex setup
  • Standards compliance - industry protocols supported natively
  • Enterprise readiness - scalable architecture for production use

References

  1. Microsoft Build 2025 Official Site
    • The official Microsoft Build conference website containing session recordings, announcements, and technical documentation. Essential for accessing the complete session content and related materials discussed in STUDIO14.
  2. Azure AI Foundry Documentation
    • Comprehensive documentation for Azure AI Foundry platform, including model catalog, agent service, and development tools. Critical for understanding the technical implementation details of concepts discussed by Yina Arenas.
  3. Azure AI Search Documentation
    • Technical documentation for Azure AI Search (formerly Azure Cognitive Search), covering indexing, retrieval, and semantic search capabilities. Essential for implementing the data foundation concepts explained by Pablo Castro.
  4. Model Context Protocol (MCP) Specification
    • Official specification for the Model Context Protocol mentioned in the session as a key standard for agent tool integration. Important for developers implementing agent systems with standardized tool calling.
  5. OpenAI Assistants API Documentation
    • Documentation for the Assistants API standard mentioned as supported by Azure AI Foundry. Relevant for understanding agent development patterns and compatibility with existing OpenAI-based systems.
  6. Azure API Management Documentation
    • Documentation for Azure API Management service demonstrated in the live demo. Important for understanding how to expose APIs in multiple protocols (OpenAPI, MCP) for agent integration.
  7. RAG (Retrieval-Augmented Generation) Research Papers
    • Academic papers on RAG methodology that provide the foundation for understanding the evolution to agentic retrieval discussed in the session. Helps contextualize the technical advancement from traditional RAG to agentic approaches.
  8. Enterprise AI Safety and Governance Guidelines
    • Microsoft’s responsible AI principles and governance framework, relevant to the safety and security discussion led by Yina Arenas. Essential for understanding enterprise-grade AI implementation requirements.
  9. LangChain Framework Documentation
    • Documentation for LangChain, mentioned as a supported framework in Azure AI Foundry. Important for developers using existing LangChain implementations who want to integrate with Azure services.
  10. BMW Group Digital Innovation Case Studies
    • BMW’s official innovation documentation that provides context for the sensor data case study discussed by Marco Casalaina. Relevant for understanding enterprise AI implementation in automotive industry.

Appendix

A.1. Technical Acronyms and Definitions

MDR (Mobile Data Recorder): BMW’s sensor system with 5,000 sensors per vehicle recording operational data including engine temperature, brake temperature, ambient conditions, and moisture levels.

MCP (Model Context Protocol): A standardized protocol for agent-to-tool communication, enabling consistent interfaces between AI models and external services or APIs.

A2A Protocol: Agent-to-Agent communication protocol for enabling coordination and collaboration between multiple AI agents in complex workflows.

Kusto: Microsoft’s data analytics service (Azure Data Explorer) used for big data analytics, mentioned in the BMW case study as their original sensor data storage solution.

A.2. Session Logistics and Format Details

The session was conducted as an intimate “STUDIO” format interview rather than a traditional presentation, allowing for natural conversation flow and deeper exploration of technical concepts. This format choice enabled the speakers to build on each other’s expertise and provide real-time clarifications and examples.

Recording Quality Note: The transcript indicates some phonetic spellings (e.g., “Descrived” likely referring to “described”) and informal conversation markers, which have been interpreted contextually in this analysis.

A.3. Humor and Personality Moments

The session included several light-hearted moments that revealed the collaborative culture of Microsoft’s AI teams:

  • Marco’s response to the “favorite color” question as “AI”
  • Seth’s creation of the “swift statement” programming analogy
  • The casual discovery that Marco had never used Azure API Management despite being at Microsoft for three years
  • The acknowledgment that API integration “may not be the sexiest thing in the world, but you need this stuff”

These moments humanized the technical discussion and demonstrated the approachable nature of the Microsoft AI leadership team.

A.4. Unaddressed Technical Limitations

While the session covered extensive capabilities, certain limitations and challenges were not explicitly discussed:

  • Cost implications of running 10,000+ model catalogs and automatic routing
  • Latency considerations in agentic retrieval systems with multiple iteration cycles
  • Data privacy and sovereignty concerns in global enterprise deployments
  • Integration complexity when working with non-Azure enterprise systems
  • Performance benchmarks comparing traditional RAG vs. agentic retrieval approaches

These topics, while important for enterprise implementation, were likely omitted due to the session’s time constraints and introductory focus.